SPLITTING FOR RARE-EVENT SIMULATION. ˆγ n has relative error. RE[ ˆγ n ] = (Var[ ˆγ n]) 1/2

Size: px
Start display at page:

Download "SPLITTING FOR RARE-EVENT SIMULATION. ˆγ n has relative error. RE[ ˆγ n ] = (Var[ ˆγ n]) 1/2"

Transcription

1 Proceedings of the 2006 Winter Simulation Conference L. F. Perrone, F. P. Wieland, J. Liu, B. G. Lawson, D. M. Nicol, and R. M. Fujimoto, eds. SPLITTING FOR RARE-EVENT SIMULATION Pierre L Ecuyer Valérie Demers Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville Montréal (Québec), H3C 3J7, CANADA Bruno Tuffin IRISA-INRIA, Campus Universitaire de Beaulieu Rennes Cedex, FRANCE ABSTRACT Splitting and importance sampling are the two primary techniques to make important rare events happen more frequently in a simulation, and obtain an unbiased estimator with much smaller variance than the standard Monte Carlo estimator. Importance sampling has been discussed and studied in several articles presented at the Winter Simulation Conference in the past. A smaller number of WSC articles have examined splitting. In this paper, we review the splitting technique and discuss some of its strengths and limitations from the practical viewpoint. We also introduce improvements in the implementation of the multilevel splitting technique. This is done in a setting where we want to estimate the probability of reaching B before reaching (or returning to) A when starting from a fixed state x 0 B, where A and B are two disjoint subsets of the state space and B is very rarely attained. This problem has several practical applications. 1 SETTING We consider a discrete-time Markov chain {X j, j 0} with state space X. Let A and B be two disjoint subsets of X and let x 0 X \ B be the initial state. The chain starts in state X 0 = x 0, leaves the set A if x 0 A, and then eventually reaches B or A. Let τ A = inf{ j > 0 : X j 1 A and X j A}, the first time when the chain hits A (or returns to A after leaving it), and τ B = inf{ j > 0 : X j B}, the first time when the chain reaches the set B. The goal is to estimate γ = P[τ B < τ A ], the probability that the chain reaches B before A. This particular form of rare-event problem, where γ is small, occurs in many practical situations (Shahabuddin 1994, Heidelberger 1995). The standard Monte Carlo method estimates γ by running n independent copies of the chain up to the stopping time τ = min(τ A,τ B ), and counting the proportion of runs for which the event {τ B < τ A } occurs. The resulting estimator ˆγ n has relative error RE[ ˆγ n ] = (Var[ ˆγ n]) 1/2 γ 0 = (γ 0(1 γ 0 )/n) 1/2 γ 0 (γ 0 n) 1/2, which increases to infinity when γ 0. This naive estimator is thus highly unreliable when γ is small. An alternative unbiased estimator of γ, say γ n, is said to have bounded relative error if lim γ 0 + RE[ γ n ] <. This implies that log(e[ γ n 2 ]) lim = 2. (1) γ 0 + logγ When the latter (weaker) condition holds, the estimator γ n is said to be asymptotically efficient (Heidelberger 1995, Bucklew 2004). To take into account the computing cost of the estimator, it is common practice to consider the efficiency of an estimator γ n of γ, defined as Eff[ γ n ] = 1/(Var[ γ n ]C( γ n )) where C( γ n ) is the expected time to compute γ n. Efficiency improvement means finding an unbiased estimator with larger efficiency than the one previously available. The estimator γ n has bounded worknormalized relative error, or relative efficiency bounded away from zero, if lim γ 0 + γ 2 Eff[ γ n ] > 0. It is worknormalized asymptotically efficient (a weaker condition) if lim γ 0 + log(c( γ n )E[ γ n 2 ])/logγ = 2. A sufficient condition for this is that (1) holds and lim γ 0 + logc( γ n )/logγ = 0. Splitting and importance sampling are the two major approaches to deal with rare-event simulation. Importance sampling increases the probability of the rare event by changing the probability laws that drive the evolution of the system. It then multiplies the estimator by an appropriate likelihood ratio to recover the correct expectation (i.e., so that the estimator remains unbiased for γ in the above setting). The main difficulty in general is to find a good way to change the probability laws. For the details, we refer the reader to Glynn and Iglehart (1989), Heidelberger (1995), Bucklew (2004), and many other references given there. In the splitting method, the probability laws remain unchanged, but an artificial drift toward the rare event is created by terminating with some probability the trajectories that seem to go away from it and by splitting (cloning) those

2 that are going in the right direction. In general, an unbiased estimator is recovered by multiplying the original estimator by an appropriate factor (in some settings, this factor is 1). The method can be traced back to Kahn and Harris (1951) and has been studied (sometimes under different names) by several authors, including Booth and Hendricks (1984), Villén-Altamirano and Villén-Altamirano (1994), Melas (1997), Garvels and Kroese (1998), Glasserman et al. (1998), Glasserman et al. (1999), Fox (1999), Garvels (2000), Del Moral (2004), Cérou, LeGland, Del Moral, and Lezaud (2005), Villén-Altamirano and Villén-Altamirano (2006), and other references cited there. The splitting methodology was invented to improve the efficiency of simulations of particle transport in nuclear physics; it is used to estimate the intensity of radiation that penetrates a shield of absorbing material, for example (Hammersley and Handscomb 1964, Spanier and Gelbard 1969, Booth and Hendricks 1984, Booth 1985, Booth and Pederson 1992, Pederson, Forster, and Booth 1997). This remains its primary area of application. It is also used to estimate delay time distributions and losses in ATM and TCP/IP telecommunication networks (Akin and Townsend 2001, Gorg and Fuss 1999). In a recent real-life application, splitting is used to estimate the probability that two airplanes get closer than a nominal separation distance, or even hit each other, in a stochastic dynamical model of air traffic where aircrafts are responsible for self-separation with each other (Blom et al. 2005). In Section 2, we review the theory and practice of splitting in a setting where we want to estimate γ = P[τ B < τ A ]. We start with multilevel splitting and then discuss more general alternatives. For multilevel splitting, we propose new variants, more efficient than the standard implementations. Numerical illustrations are given in Section 3. L Ecuyer, Demers, and Tuffin (2006) contains an expanded version of the present overview article. It also studies the combination of splitting and importance sampling with two types of randomized quasi-monte Carlo methods: the classical one (e.g., Owen 1998, L Ecuyer and Lemieux 2000) and the array-rqmc method for Markov chains proposed by L Ecuyer, Lécot, and Tuffin (2005). 2 SPLITTING 2.1 Multilevel Splitting We define the splitting algorithm via an importance function h : X R that assigns a importance value to each state of the chain (Garvels, Kroese, and Van Ommeren 2002). We assume that A = {x X : h(x) 0} and B = {x X : h(x) l} for some constant l > 0. In the multilevel splitting method, we partition the interval [0, l) in m subintervals with boundaries 0 = l 0 < l 1 < < l m = l. For k = 1,...,m, let T k = inf{ j > 0 : h(x j ) l k }, let D k = {T k < τ A } denote the event that h(x j ) reaches level l k before reaching level 0, and define the conditional probabilities p k = P[D k D k 1 ] for k > 1, and p 1 = P[D 1 ]. Since D m D m 1 D 1, we have γ = P[D m ] = m k=1 The intuitive idea of multilevel splitting is to estimate each probability p k separately, by starting a large number of chains in states that are generated from the distribution of X Tk 1 conditional on the event D k 1. This conditional distribution, denoted by G k 1, is called the (first-time) entrance distribution at threshold l k 1, for k = 1,...,m + 1 (G 0 is degenerate at x 0 ). Conceptually, the estimation is done in successive stages, as follows. In the first stage, we start N 0 independent chains from the initial state x 0 and simulate each of them until time min(τ A, T 1 ). Let R 1 be the number of those chains for which D 1 occurs. Then ˆp 1 = R 1 /N 0 is an obvious unbiased estimator of p 1. The empirical distribution Ĝ 1 of these R 1 entrance states X T1 can be viewed as an estimate of the conditional distribution G 1. In stage k, for k 2, ideally we would like to generate N k 1 states independently from the entrance distribution G k 1. Or even better, to generate a stratified sample from G k 1. But we usually cannot do that, because G k 1 is unknown. Instead, we pick N k 1 states out of the R k 1 that are available (by cloning if necessary), simulate independently from these states up to time min(τ A, T k ), and estimate p k by ˆp k = R k /N k 1 where R k is the number of chains for which D k occurs. If R k = 0, then ˆp j = 0 for all j k and the algorithm can immediately return ˆγ n = 0. The initial state of each of the N k 1 chains at the beginning of stage k has distribution G k 1. Thus, for each of these chains, the event D k has probability p k and the entrance state at the next level if D k occurs has distribution G k. Even though the ˆp k s are not independent, we can prove by induction on k that the product ˆp 1 ˆp m = (R 1 /N 0 )(R 2 /N 1 ) (R m /N m 1 ) is an unbiased estimator of γ (Garvels 2000, page 17): If we assume that E[ ˆp 1 ˆp k 1 ] = p 1 p k 1, then p k. E[ ˆp 1 ˆp k ] = E[ ˆp 1 ˆp k 1 E[ ˆp k N 0,R 1,...,N k 1 ]] = E[ ˆp 1 ˆp k 1 (N k 1 p k )/N k 1 ] = p 1 p k. Combining this with the fact that E[ ˆp 1 ] = p 1, the result follows. 2.2 Fixed Splitting vs Fixed Effort There are many ways of doing the splitting (Garvels 2000). For example, we may clone each of the R k chains that reached level k in c k copies, for a fixed positive integer c k. Then, each N k = c k R k is random. This is fixed splitting. If

3 we want the expected number of splits of each chain to be c k, where c k = c k + δ and 0 δ < 1, then we assume that the actual number of splits is c k +1 with probability δ and c k with probability 1 δ. In the fixed effort method, we fix each N k a priori and make just the right amount of splitting to reach this target value. This can be achieved by random assignment: draw the N k starting states at random, with replacement, from the R k available states. This is equivalent to sampling N k states from the empirical distribution Ĝ k of these R k states. In a fixed assignment, on the other hand, we split each of the R k states approximately the same number of times as follows. Let c k = N k /R k and d k = N k mod R k. Select d k of the R k states at random, without replacement. Each selected state is split c k + 1 times and the other states are split c k times. The fixed assignment gives a smaller variance than the random assignment because it corresponds to stratification over the empirical distribution Ĝ k at level k. These variants are all unbiased, but they differ in terms of variance. Garvels and Kroese (1998) conclude from their analysis and empirical experiments that fixed effort performs better, mainly because it reduces the variance of the number of chains that are simulated at each stage. It turns out that with optimal splitting factors, this is not always true (see the next subsection). The fixed effort implementation with random assignment fits the framework of interacting particle systems studied by Del Moral (2004) to approximate Feynman-Kac distributions. In this type of system, particles that did not reach the threshold are killed and replaced by clones of randomly selected particles among those that have succeeded. This redistributes the effort on most promising particles while keeping the total number constant. Cérou, LeGland, Del Moral, and Lezaud (2005) derive limit theorems for the corresponding estimators. 2.3 Variance Analysis for a Simplified Setting We outline a very crude variance analysis in an idealized fixed-effort setting where N 0 = N 1 = = N m 1 = n and where the ˆp i s are independent binomial random variables with parameters n and p = γ 1/m. Then, for m > 1, we have (Garvels 2000, 2006): = = Var[ ˆp 1 ˆp m ] m i=1 E[ ˆp 2 i ] γ 2 ( p 2 + ) p(1 p) m p 2m n = mp2m 1 (1 p) + m(m 1)p2m 2 (1 p) 2 n 2n 2 (p(1 p))m + + n m. If we assume that n (m 1)(1 p)/p, (2) the first term mp 2m 1 (1 p)/n mγ 2 1/m /n dominates in the last expression. The standard Monte Carlo variance, on the other hand, is γ(1 γ)/n γ/n. To illustrate the huge potential variance reduction, suppose γ = 10 20, m = 20, p = 1/10, and n = Then the MC variance is whereas mp 2m 1 (1 p)/n This oversimplified setting is not realistic, because the ˆp i are generally not independent and it is difficult to have p i = γ 1/m for all i, but it gives an idea of the order of magnitude of potential variance reduction. The amount of work (or CPU time, or number of steps simulated) at each stage is proportional to n, so the total work is proportional to nm. Most of this work is to simulate the n chains down to level 0 at each stage. Thus, the efficiency of the splitting estimator under the simplified setting is approximately proportional to n/[γ 2 1/m nm 2 ] = γ 2+1/m /m 2 when (2) holds. By differentiating with respect to m, we find that this expression is maximized by taking m = ln(γ)/2 (we neglect the fact that m must be an integer). This gives p m = γ = e 2m, so p = e 2. Garvels and Kroese (1998) have obtained this result. The squared relative error in this case is (approximately) γ 2 1/m (m/n)γ 2 = e 2 m/n = e 2 ln(γ)/(2n) and the relative efficiency is proportional to γ 2 γ 2+1/m /m 2 = (em) 2 = [(e/2)ln(γ)] 2, again under the condition (2). When γ 0 for fixed p, we have m, so (2) does not hold. Then, the relative error increases toward infinity and the relative efficiency converges to zero, at a logarithmic rate in both cases. This agrees with Garvels (2000), page 20. With γ n = ˆp 1 ˆp m, the limit in (1) is log(p 2 + p(1 p)/n) m lim γ 0 + logγ = lim γ 0 + log(p 2 + p(1 p)/n) log p < 2. Thus, this splitting estimator is not quite asymptotically efficient, but almost (when n is very large).

4 Consider now a fixed-splitting setting, assuming that N 0 = n, p k = p = γ 1/m for all k, and that the constant splitting factor at each stage is c = 1/p; i.e., N k = R k /p. Then, {N k, k 1} is a branching process and the estimator becomes ˆp 1 ˆp m = R 1 R 2 N 0 N 1 R m = R m p m 1 N m 1 n From standard branching process theory (Harris 1963), we have that Var[ ˆp 1 ˆp m ] = m(1 p)p 2m 1 /n. If p is fixed and m, then the squared relative error m(1 p)/(np) is unbounded here as well. However, the limit in (1) becomes log(m(1 p)γ 2 /(np) + γ 2 ) lim γ 0 + logγ = lim γ 0 + 2mlog p log(1 + m(1 p)/(np)) mlog p. = 2, so the splitting estimator is asymptotically efficient (Glasserman et al. 1999). This implies that fixed splitting is asymptotically better in this case. Glasserman et al. (1999) study the fixed splitting framework with splitting factor c k c, for a countable-state space Markov chain. They assume that the probability transition matrix P k for the first-entrance state at level k given the first-entrance state at level k 1 converges to a matrix P with spectral radius ρ < 1. This implies that p k ρ when k. Then they use branching process theory to prove that the multilevel splitting estimator (in their setting) is worknormalized asymptotically efficient if and only if c = 1/ρ. Glasserman et al. (1998) show that the condition c = 1/ρ is not sufficient for asymptotic efficiency and provide additional necessary conditions in a general multidimensional setting. Their results highlight the crucial importance of choosing a good importance function h. Even though fixed splitting is asymptotically better under ideal conditions, its efficiency is extremely sensitive to the choice of splitting factors. If the splitting factors are too high, the number of chains (and the amount of work) explodes, whereas if they are too low, the variance is very large because very few chains reach B. Since the optimal splitting factors are unknown in real-life applications, the more robust fixed-effort approach is usually preferable. 2.4 Implementation The fixed-effort approach has the disadvantage of requiring more memory than fixed splitting, because it must use a breadth-first implementation: at each stage k all the chains must be simulated until they reach either A or level l k before we know the splitting factor at that level. The states of all the chains that reach l k must be saved; this may require too much memory when the N k s are large. With fixed splitting, we can adopt a depth-first strategy, where each chain is simulated entirely until it hits l or A, then its most recent clones (created at the highest level that it has reached) are simulated entirely, then those at the next highest level, and so on. This procedure is applied recursively. At most one state per level need to be memorized with this approach. This is feasible because the amount of splitting at each level is fixed a priori. As a second issue, an important part of the work in multilevel splitting is due to the fact that all the chains considered in stage k (from level l k 1 ) and which do not reach l k must be simulated until they get down to A. When l k 1 is large, this can take significant time. Because of this, the expected amount of work increases with the number of thresholds. One heuristic that reduces this work in exchange for a small bias truncates the chains that reach level l k β downward after they have reached l k 1, where β 2 is a fixed integer large enough so that a chain starting at level l k β has a very small probability of getting back up to l k. We discuss unbiased alternatives in Section The RESTART Algorithm The RESTART method (Villén-Altamirano and Villén-Altamirano 1994, Villén-Altamirano and Villén-Altamirano 2006) is a variant of splitting where any chain is split by a fixed factor when it hits a level upward, and one of the copies is tagged as the original for that level. When any of those copies hits that same level downward, if it is the original it just continues its path, otherwise it is killed immediately. This rule applies recursively, and the method is implemented in a depth-first fashion, as follows: whenever there is a split, all the non-original copies are simulated completely, one after the other, then simulation continues for the original chain. Unbiasedness is proved by Garvels (2000) and Villén-Altamirano and Villén-Altamirano (2002). The reason for killing most of the paths that go downward is to reduce the work. The number of paths that are simulated down to A never exceeds N 0. On the other hand, the number of chains that reach a given level is more variable with this method than with the fixed-effort and fixed-assignment multilevel splitting algorithm described previously. As a result, the final estimator of γ has a larger variance (Garvels 2000). Another source of additional variance is that the resplits tend to share a longer common history and to be more positively correlated. This source of variance can be important when the probability of reaching B from a given level varies significantly with the entrance state at that level (Garvels 2000). In terms of overall efficiency, none of the two methods is universally better; RESTART wins in some cases and splitting wins in other cases. Villén-Altamirano and Villén-Altamirano (2002) provide a detailed variance analysis of RESTART.

5 2.6 Choice of the Importance Function and Optimal Parameters Key issues in multilevel splitting are the choices of the importance function h, levels l k, and splitting factors. To discuss this, we introduce some more notation. Let X k X be the support of the entrance distribution G k, i.e., the states in which the chain can possibly be when hitting level l k for the first time. Let γ(x) = P[τ B < τ A τ > j, X j = x], the probability of reaching B before A if the chain is currently in state x, and p k (x) = P[D k D k 1,X Tk 1 = x], the probability of reaching level k before hitting A if the chain has just entered level k 1 in state x, for x X k 1. Note that p k = x X k 1 p k (x)dg k 1 (x) and γ = γ(x 0 ). One-dimensional case: Selecting the levels. If the Markov chain has a one-dimensional state space X R, γ(x) is increasing in x, and if A = (,0] and B = [l, ) for some constant l, then we could simply choose h(x) = x (or any strictly increasing function). In this case, the kth level is attained when the state reaches the value l k. This value need not be reached exactly: in general, the chain can jump directly from a smaller value to a value larger than l k, perhaps even larger than l k+1. So even in the one-dimensional case, the entrance state x at a given level is not unique in general and the probability p k (x) of reaching the next level depends on this (random) entrance state. It remains to choose the levels l k. We saw earlier that in a fixed effort setting and under simplifying assumptions, it is optimal to have p k p = e 2 for all k. This gives m = ln(γ)/2 levels. To obtain equal p k s, it is typically necessary to take unequal distances between the successive levels l k, i.e., l k l k 1 must depend on k. Suppose now that we use fixed splitting with c k = 1/p k = e 2 for each k. If we assume (crudely) that each chain is split by a factor of e 2 at each stage, the total number of copies of a single initial chain that have a chance to reach B is e 2m 2 = e ln(γ) 2 = e 2 γ 1. (3) Since each one reaches B with probability γ, this crude argument indicates that the expected number of chains that reach B is approximately equal to p = e 2 times the initial number of chains at stage 0, exactly as in the fixed-effort case. However, the variance generally differs. For RESTART, Villén-Altamirano and Villén-Altamirano (1994) concluded from a crude analysis that p k e 2 was approximately optimal. However, their more careful analysis in Villén-Altamirano and Villén-Altamirano (2002) indicates that the p k s should be as small as possible. Since the splitting factor at each level must be an integer, they recommend p k = 1/2 and a splitting factor of c k = 2. Cérou and Guyader (2005) determine the thresholds adaptively for the splitting with fixed effort in dimension 1. They first simulate n chains (trajectories) until these chains reach A or B. Then they sort the chains according to the maximum value of the importance function h that each chain has reached. The k trajectories with the largest values are kept, while the n k others are re-simulated, starting from the state at which the highest value of the importance function was obtained for the (n k)-th largest one. They proceed like this until n k trajectories have reached B. Their estimator is proved to be consistent, but is biased. Multidimensional case: Defining the importance function. In the case of a multidimensional state space, the choice of h is much more difficult. Note that h and the l k s jointly determine the probabilities p k (x) and p k. Based on large deviation theory, Glasserman et al. (1998) shows that the levels need to be chosen in a way consistent with the most likely path to a rare set. Garvels, Kroese, and Van Ommeren (2002) show by induction on k that for any fixed p 1,..., p m, h should be defined so that p k (x) = p k (independent of x) for all x X k 1 and all k. This rule minimizes the residual variance of the estimator from stage k onward. With an h that satisfies this condition, the optimal levels and splitting factors are the same as in the one-dimensional case: m = (1/2)lnγ levels, p k e 2 and E[N k ] = N 0 for each k. A simple choice of h and l k s that satisfies these conditions is h(x) = h (x) def = γ(x) and l k = e 2(m k) = γe 2k. Garvels, Kroese, and Van Ommeren (2002) gave the following (equivalent) alternative choice: l k = k for each k and h(x) = h (x) def = ln(γ(x)/γ) = m + ln(γ(x)) 2 2 for all x X. However, these levels are optimal only if we assume that the chain can reach l k only on the set {x : γ(x) = e 2(m k) }, an optimistic assumption that rarely holds in practice, especially in the multidimensional case. Garvels, Kroese, and Van Ommeren (2002) also show how to get a first estimate of γ(x) beforehand, in simple situations where the Markov chain has a finite state space, by simulating the chain backward in time. They construct an approximation of h from this estimate and then use it in their splitting algorithm. They apply their method to a tandem queue with two or three nodes and obtain good results. However, this method appears to have limited applicability for large and complicated models. Booth and Hendricks (1984) propose adaptive methods that learn the importance function as follows. In their setting, the state space is partitioned in a finite number of regions and the importance function h is assumed to be constant in each region. This importance function is used to determine the expected splitting factors and Russian roulette probabilities (see Section 2.8) when a chain jumps from one region to another. They estimate the average value of γ(x)

6 in each region by the fraction of chains that reach B among those that have entered this region. These estimates are taken as importance functions in further simulations used to improve the estimates, and so on. Constructing the functions h or h essentially requires the knowledge of the probability γ(x) for all x. But if we knew these probabilities, there would be no need for simulation! This is very similar (and related) to the issue of constructing the optimal change of measure in importance sampling (Glasserman et al. 1998). In general, finding an optimal h, or an h for which p k (x) is independent of x, can be extremely difficult or even impossible. When p k (x) depends on x, selecting the thresholds so that p k e 2 is not necessarily optimal. More importantly, with a bad choice of h, splitting may increase the variance, as illustrated by the next example. Example 1 This example was used by Parekh and Walrand (1989), Glasserman et al. (1998), Glasserman et al. (1999), and Garvels (2000), among others. Consider an open tandem Jackson network with two queues, arrival rate 1, and service rate µ j at queue j for j = 1,2. Let X j = (X 1, j, X 2, j ) denote the number of customers at each of the two queues immediately after the jth event (arrival or end of service). We have A = {(0,0)} and B = {(x 1,x 2 ) : x 2 l } for some large integer l. A naive choice of importance function here would be h(x 1,x 2 ) = x 2. This seems natural at first sight because the set B is defined in terms of x 2 only. With this choice, the entrance distribution at level k turns out to be concentrated on pairs (x 1,x 2 ) with small values of x 1. To see why, suppose that x 2 = l k > 0 for some integer k and that we are in state (x 1,x 2 1) where x 1 > 0 is small. The possible transitions are to states (x 1 +1,x 2 1), (x 1,x 2 2), and (x 1 1,x 2 ), with probabilities proportional to 1, µ 2, and µ 1, respectively. But the chains that go to state (x 1 1,x 2 ) are cloned whereas the other ones are not, and this tends to increase the population of chains with a small x 1. Suppose now that µ 1 < µ 2 (the first queue is the bottleneck). In this case, the most likely paths to overflow are those where the first queue builds up to a large level and then the second queue builds up from the transfer of customers from the first queue (Heidelberger 1995). The importance function h(x 1,x 2 ) = x 2 does not favor these types of paths; it rather favors the paths where x 1 remains small and these paths have a very high likelihood of returning to (0,0) before overflow. As a result, splitting with this h may give an even larger variance than no splitting at all. For this particular example, h increases in both x 1 and x 2 (Garvels, Kroese, and Van Ommeren 2002). 2.7 Unbiased Truncation We pointed out earlier that a large fraction of the work in multilevel splitting is to simulate the chains down to level zero at each stage. Truncating the chains whenever they fall below some level l k β in stage k reduces the work but introduces a bias. A large β may give negligible bias, but also a small work reduction. In what follows, we describe unbiased truncation techniques based on the Russian roulette principle (Kahn and Harris 1951, Hammersley and Handscomb 1964). Probabilistic truncation. The idea here is to kill the chains at random, with some probability, independently of each other. The survivors act as representatives of the killed chains. For stage k, we select real numbers r k,2,...,r k,k 1 in [1, ). The first time a chain reaches level l k j from above during that stage, for j 2, it is killed with probability 1 1/r k, j. If it survives, its weight is multiplied by r k, j. (This is a version of Russian roulette.) When a chain of weight w > 1 reaches level l k, it is cloned into w additional copies with probability δ = w w and w 1 additional copies with probability 1 δ. Each copy is given weight 1. Now, the number of representatives retained at any given stage is random. Note that we may have r k, j = 1 for some values of j. Periodic truncation. To reduce the variability of the number of selected representatives at each level l k j, we may decide to retain every r k, j -th chain that down-crosses that level and multiply its weight by r k, j ; e.g., if r k, j = 3, we keep the third, sixth, ninth, etc. This would generally give a biased estimator, because the probability that a chain is killed would then depend on its sample path up to the time when it crosses the level (for instance, the first chain that down-crosses the level would always be killed if r k, j > 1). A simple trick to remove that bias is to modify the method as follows: generate a random integer D k, j uniformly in {1,...,r k, j }, retain the (ir k, j + D k, j )-th chain that down-crosses level l k j for i = 0,1,2,..., and kill the other ones. We assume that the random variables D k,2,...,d k,k 1 are independent. Then, any chain that down-crosses the level has the same probability 1 1/r k, j of being killed, independently of its trajectory above that level. This is true for any positive integer r k, j. Moreover, the proportion of chains that survive has less variance than for the probabilistic truncation (the killing indicators are no longer independent across the chains). The chains that reach l k are cloned in proportion to their weight, exactly as in the probabilistic truncation. Tag-based truncation. In the periodic truncation method, the level at which a chain is killed is determined only when the chain reaches that level. An alternative is to fix all these levels right at the beginning of the stage. We first select positive integers r k,2,...,r k,k 1. Then each chain is tagged to the level l k j with probability q k, j = (r k, j 1)/(r k,2 r k, j ) for j = 2,...,k 1, and to level l 0 with probability 1 q k,k 1 q k,2 = 1/(r k,2 r k,k 1 ). Thus, all the chains

7 have the same probability of receiving any given level and the probability of receiving level zero is positive. If the tags are assigned randomly and independently across the chains, then this method is equivalent to probabilistic truncation. But if the integers r k,2,...,r k,k 1 are chosen so that their product divides (or equals) N k, the number of chains at the beginning of stage k, then the tags can also be assigned so that the proportion of chains tagged to level l k j is exactly q k, j, while the probability of receiving a given tag is the same for all chains. The reader can verify that the following scheme gives one way of achieving this: Put the N k chains in a list (in any order), generate a random integer D uniformly in {0,...,N k 1}, and assign the tag k j (i,d) to the i-th chain in the list, for all i, where j (i,d) is the smallest integer j in {2,...,k} such that r k,2 r k, j does not divide (D + i) mod N k (when (D + i) mod N k = 0, we put j (i,d) = k).. After the tags are assigned, the chains can be simulated one by one for that stage. Whenever a chain down-crosses for the first time (in this stage) a level l k j higher than its tag, its weight is multiplied by r k, j. If it down-crosses the level of its tag, it is killed immediately. The chains that reach l k are cloned in proportion to their weight, as before. Unbiasedness. (2006) show that all the above truncation methods are unbiased by proving the next proposition and then showing that each truncation method satisfies the assumptions of the proposition. Proposition 1 Suppose there are real numbers r k,2,...,r k,k 1 in [1, ) such that for j = 2,...,k 1, each chain has a probability 1 1/r k, j of being killed at its first down-crossing of level l k j, independently of its sample path up to that moment, and its weight is multiplied by r k, j if it survives. Then the truncated estimator remains unbiased. Getting rid of the weights. In the unbiased truncation methods discussed so far, the surviving chains have different weights. The variance of these weights may contribute significantly to the variance of the final estimator. For example, if k is large, the event that a chain reaches l k (from l k 1 ) after going down to l 1 is usually a rare event, and when it occurs the corresponding chain has a large weight, so this may have a non-negligible impact on the variance. This can be addressed by resplitting the chains within the stage when they up-cross some levels, instead of increasing their weights at down-crossings. We explain how the probabilistic and tag-based truncation methods can be modified to incorporate this idea. In these methods, the weights of all chains are always 1, and whenever a chain down-crosses l k j (not only the first time), for j 2, it can get killed. Probabilistic truncation and resplitting within each stage. The probabilistic truncation method can be modified as follows. During stage k, whenever a chain reaches a level l k j from below, it is split in r k, j identical copies that start evolving independently from that point onward (if r k, j is not an integer, we split the chain in r k, j + 1 copies with probability δ = r k, j r k, j and in r k, j copies with probability 1 δ). Whenever a chain down-crosses l k j (not only the first time), for j 2, it is killed with probability 1 1/r k, j. All chains always have weight 1. Tag-based truncation with resplits. This method is equivalent to applying RESTART separately within each stage of the multistage splitting algorithm. It modifies the tag-based truncation as follows: Whenever a chain upcrosses level l k j for j 2, it is split in r k, j copies. One of these r k, j copies is identified as the original and keeps its current tag, while the other r k, j 1 copies are tagged to the level l k j where the split occurs. As before, a chain is killed when it down-crosses the level of its tag. Unbiasedness. (2006) prove the following proposition and show that the two truncation methods with resplits that we just described satisfy its assumptions. Proposition 2 Suppose there are positive real numbers r k,2,...,r k,k 1 such that for j = 2,...,k 1, each chain is killed with probability 1 1/r k, j whenever it down-crosses level l k j, independently of its sample path up to the time when it reached that level, and that this chain is split into C chains when it up-crosses that same level, where C is a random variable with mean r k, j, independent of the history so far. Then the estimator with probabilistic truncation and resplits (without weights) is unbiased for γ. Effectiveness and Implementation. The resplit versions of the truncation methods are expected to give a smaller variance but require more work. So there is no universal winner if we think of maximizing the efficiency. One disadvantage of the resplit versions is that the number of chains alive at any given time during stage k has more variance and may exceed N k 1. In a worst-case situation, a chain may go down and up many times across several levels without being killed, giving rise to a flurry of siblings along the way. Fortunately, this type of bad behavior has an extremely small probability and poses no problem when the splitting parameters are well chosen. In all our experiments, the number of chains alive simultaneously during any given stage k has rarely exceeded N k 1. If we want to insist that the number of chains never exceeds N k 1, we can use weights instead of splitting, but just for the splits that would have made the number of chains too large. We may want to do that if the chains are stored in an array of size n = N k 1 and we do not

8 want their number to exceed n. This type of implementation is needed when we combine splitting with the array-rqmc method ( 2006). We have a lot of freedom for the choice of the truncation and resplit parameters r k, j. We can select different sets of values at the different stages of the multilevel splitting algorithm. It appears sensible to take r k, j = 1/ ˆp k j = N k j 1 /R k j, the actual splitting factor used at level l k j of the splitting algorithm, for j 2. In our experiments, this has always worked well. split in C copies, the weight of all the copies is set to either w/c or w/e[c]. Booth (1985) shows that using w/e[c] is usually better. When Russian roulette is applied, the chain is killed with some probability α < 1; if it survives, its weight is multiplied by 1/(1 α). The values of C and α at each step can be deterministic or random, and may depend on the past history of the chain. Whenever a cost is incurred, it must be multiplied by the weight of the chain. Unbiasedness for this general setting is proved (under mild conditions) by Booth and Pederson (1992), for example. 2.8 Getting Rid of the Levels In some versions of the splitting and Russian roulette technique, there are no levels (or thresholds), but only an importance function (some authors call it branching function). For instance, Ermakov and Melas (1995) and Melas (1997) study a general setting where a chain can be split or killed at any transition. If the transition is from x to y and if α = h(y)/h(x) 1, then the chain is split in a random number C of copies where E[C] = α, whereas if α < 1 it is killed with probability 1 α (this is Russian roulette). In case of a split, the C 1 new copies are started from state x and new transitions are generated (independently) for those chains. Their method is developed to estimate the average cost per unit of time in a regenerative process, where a statedependent cost is incurred at each step. In the simulation, each cost incurred in a given state x is divided by h(x). We may view 1/h(x) as the weight of the chain at that point. At the end of a regenerative cycle, the total weighted cost accumulated by the chain over its cycle is the observation associated with this cycle. The expected cost per cycle is estimated by averaging the observations over all simulated cycles. The expected length of a cycle is estimated in the same way, just replacing costs by lengths. The authors show that their method is consistent and propose an adaptive algorithm that estimates the optimal h. This method can be applied to a finite-horizon simulation as well. In our setting, it suffices to replace the regeneration time by the time when the chain reaches A or B, and then forget about the length of the cycle. When a chain reaches B, it contributes its weight 1/h(X τb ) to the estimator. For a very crude analysis, suppose we take h(x) = γ(x) and that there is a split in two every time the function h doubles its value. Here, h(y)/h(x) = γ(y)/γ(x), so a chain that reaches the set B would have split in two approximately log 2 γ times. This gives a potential of 2 log 2 γ = 1/γ copies that can possibly reach B for each initial chain at level 0, the same number as for the multilevel splitting and RESTART; see Equation (3). This argument suggests that an optimal h in this case should be proportional to γ(x). In general, splitting and Russian roulette can be implemented by maintaining a weight for each chain. Initially, each chain has weight 1. Whenever a chain of weight w is 2.9 Weight Windows Particle transport simulations in nuclear physics often combine splitting and Russian roulette with importance sampling. Then, the weight of each chain must be multiplied by the likelihood ratio accumulated so far. The weight is redefined as this product. In the context of rare events, it is frequently the case that the final weight of a chain is occasionally large and usually very small. This gives rise to a large variance and a highly-skewed distribution, for which variance estimation is difficult. To reduce the variance of the weights, Booth (1982) introduced the idea of weight windows, which we define as follows (see also Booth and Hendricks (1984) and Fox (1999)). Define the weighted importance of a chain as the product of its weight w and the value of the importance function h(x) at its current state. Select three real numbers 0 < a min < a < a max. Whenever the weighted importance ω = wh(x) of a chain falls below a min, we apply Russian roulette, killing the chain with probability 1 ω/a. If the chain survives, its weight is set to a/h(x). If the weighted importance ω rises above a max, we split the chain in c = ω/a max copies and give weight w/c to each copy. The estimator of γ = P[τ B < τ A ] is the sum of weights of all the chains that reach the set B before reaching A. The importance function h (x) = γ(x) should be approximately optimal in this case. The basic motivation is simple: if the weight window is reasonably narrow, all the chains that reach B would have approximately the same weight, so the only significant source of variance would be the number of chains that reach B (Booth and Hendricks 1984). If we take a = (a min + a max )/2 γ, then this number has expectation n (approximately), where n is the initial number of chains. In the original proposal of Booth (1982) and Booth and Hendricks (1984), the windows are on the weights, not on the weighted importance. The state space is partitioned in a finite number of regions (say, up to 100 regions), the importance function is assumed constant in each region, and each region has a different weight window, inversely proportional to the value of the importance function in that region. Such weight windows are used extensively in the Los Alamos particle transport simulation programs. Our formulation is essentially equivalent, except that we do not assume a finite

9 partition of the state space. Fox (1999), Chapter 10) discusses the use of weight windows for splitting and Russian roulette, but does not mention the use of an importance function. Weight windows without an importance function could be fine when a good change of measure (importance sampling) is already applied to drive the system toward the set B. Then, the role of splitting and Russian roulette is only to equalize the contributions of the chains that reach B and kill most of those whose anticipated contribution is deemed negligible, to save work. This type of splitting, based only on weights and without an importance function, gives no special encouragement to the chains that go toward B. If we use it alone, the event {τ B < τ A } will remain a rare event. If there is no importance sampling, the multilevel splitting techniques described earlier (except those with truncation and no resplits, in Section 2.7) have the advantage of not requiring explicit (random) weights. All the chains that reach level l k have the same weight when they reach that level for the first time. So there is no need for weight windows in that context. 3 EXAMPLES Example 2 We return to Example 1, an open tandem Jackson queueing network with two queues. The choice of h is crucial for this example, especially if µ 1 < µ 2 (Glasserman et al. 1998). Here we look at a case where µ 1 > µ 2. We consider the following choices of h: h 1 (x 1,x 2 ) = x 2 ; (4) h 2 (x 1,x 2 ) = (x 2 + min(0,x 2 + x 1 l))/2; (5) h 3 (x 1,x 2 ) = x 2 + min(x 1,l x 2 1) (1 x 2 /l).(6) The function h 1 is a naive choice based on the idea that the set B is defined in terms of x 2 only. The second choice, h 2, counts l minus half the minimal number of steps required to reach B from the current state. (To reach B, we need at least l min(0,x 2 + x 1 l) arrivals at the first queue and l x 2 transfers to the second queue.) The third choice, h 3, is adapted from Villén-Altamirano (2006), who recommends h(x 1,x 2 ) = x 2 + x 1 when µ 1 > µ 2. This h was modified as follows. We define h 3 (x) = x 1 +x 2 when x 1 +x 2 l 1 and h 3 (x) = l when x 2 l. In between, i.e., in the area where l x 1 1 x 2 l, we interpolate linearly in x 2 for any fixed x 1. This gives h 3. We did a numerical experiment with µ 1 = 4, µ 2 = 2, and l = 30, with our three choices of h. For each h and each truncation method discussed earlier, we computed the variance per chain, V n = nvar[ ˆγ n ], where n is the (expected) number of chains at each level, and the work-normalized variance per chain, W n = S n Var[ ˆγ n ], where S n is the expected total number of simulated steps of the n Markov chains. If S n is seen as the computing cost of the estimator, then 1/W n is the usual measure of efficiency. For fixed splitting without truncation and resplits, V n and W n do not depend on n. Here we briefly summarize the detailed results given in (2006). We have V n γ with standard Monte Carlo (no splitting) and V n with the multilevel splitting with h 2, using fixed effort and no truncation. This is a huge variance reduction. With h 1, ˆV n and Ŵ n were significantly higher than for h 2 and h 3, whereas h 3 was just a bit better than h 2. The truncation and resplit methods improved the efficiency roughly by a factor of 3. There is slightly more variance reduction with the variants that use resplits than with those that do not resplit, but also slightly more work, and the efficiency remains about the same. Example 3 We consider an Ornstein-Uhlenbeck stochastic process {R(t), t 0}, which obeys the stochastic differential equation dr(t) = a(b R(t))dt + σdw(t) where a > 0, b, and σ > 0 are constants, and {W(t), t 0} is a standard Brownian motion (Taylor and Karlin 1998). This is the Vasicek model for the evolution of short-term interest rates (Vasicek 1977). In that context, b can be viewed as a long-term interest rate level toward which the process is attracted with strength a(b R(t)). Suppose the process is observed at times t j = jδ for j = 0,1,... and let X j = R(t j ). Let A = (,b], B = [l, ) for some constant l, and x 0 b. We want to estimate the probability that the process exceeds level l at one of the observation times before it returns below b, when started from R(0) = x 0. Here we take b = 0. Suppose we take the importance function h equal to the identity. The thresholds l k should be placed closer to each other as k increases, because the attraction toward b = 0 becomes stronger. Preliminary empirical experiments suggest the following rule, which makes the p k s approximately independent of k: set tentatively l k = l k/m for k = 1,...,m, let k be the largest k for which l k < 2, and reset l k = l k (k/k ) for k = 1,...,k 1. The latter makes the first thresholds approximately equidistant. Because of the time discretization, the entrance distribution G k has positive support over the entire interval [l k, ). This means that a chain can cross an arbitrary number of thresholds in a single jump. The simulation starts from a fixed state only at the first level. We made some experiments with a = 0.1, b = 0, σ = 0.3, x 0 = 0.1, δ = 0.1, l = 4, and m = 14 levels. With these parameters, we have V n γ with standard Monte Carlo (no splitting) and V n with the multilevel splitting without truncation, with either the fixed splitting or fixed effort approach. The truncation and resplit methods improve the work-normalized variance W n roughly by a factor of 3, as in the previous example. The work is reduced by

10 a factor of 4.3 without the resplits and by a factor of 3.5 with the resplits, but the variance is increased roughly by a factor of 1.4 without the resplits and 1.2 with the resplits. The benefits of splitting and of truncation increase with l. For l = 6, for example, we have V n γ with standard Monte Carlo and V n with the multilevel splitting without truncation, with m = 30 (this gives p k s of approximately the same size as with l = 4 and m = 14). In this case, the truncation and resplit methods reduce the work-normalized variance approximately by a factor of 8 to 10. Fixed effort and fixed splitting also have comparable efficiencies when no truncation is used. Example 4 There are situations where the splitting method is not appropriate whereas importance sampling can be made very effective. Consider for example a highly-reliable Markovian multicomponent system (Shahabuddin 1994) for which the failure of a few components (e.g., 2 or 3) may be sufficient for the entire system to fail, and where all the components have a very small failure rate and a high repair rate. If we want to apply splitting, the thresholds must be defined in terms of the vector of failed components (the state of the system). But whenever there are failed components, the next event is a repair with a very high probability. So regardless of how we determine the thresholds, the probabilities p k of reaching the next threshold from the current one are always very small. For this reason, the splitting method cannot be made efficient in this case. On the other hand, there are effective importance sampling methods for this type of model (Shahabuddin 1994, Cancela, Rubino, and Tuffin 2002). 4 CONCLUSION Splitting is a valuable but seemingly under-exploited variance reduction technique for rare-event simulation. It certainly deserves further study. In multidimensional settings, finding out an appropriate importance function h can be a difficult task and seems to the the main bottleneck for an effective application of the method. Providing further hints in this direction, and developing adaptive techniques to learn good importance functions, would be of significant interest. Unfortunately, splitting can hardly be applied to problems where rarity comes from the occurrence of a lowprobability transition that cannot be decomposed in several higher-probability transitions. ACKNOWLEDGMENTS This research has been supported by NSERC-Canada grant No. ODGP and a Canada Research Chair to the first author, an NSERC-Canada scholarship to the second author, and EuroNGI Network of Excellence, INRIA s cooperative research initiative RARE and SurePath ACI Security Project to the third author, and an FQRNT-INRIA Travel Grant to the first and third authors. REFERENCES Akin, O., and J. K. Townsend Efficient simulation of TCP/IP networks characterized by non-rare events using DPR-based splitting. In Proceedings of IEEE Globecom, Blom, H. A. P., G. J. Bakker, J. Krystul, M. H. C. Everdij, B. K. Obbink, and M. B. Klompstra Sequential Monte Carlo simulation of collision risk in free flight air traffic. Technical report, Project HYBRIDGE IST Booth, T. E Automatic importance estimation in forward Monte Carlo calculations. Transactions of the American Nuclear Society 41: Booth, T. E Monte Carlo variance comparison for expected-value versus sampled splitting. Nuclear Science and Engineering 89: Booth, T. E., and J. S. Hendricks Importance estimation in forward Monte Carlo estimation. Nuclear Technology/Fusion 5: Booth, T. E., and S. P. Pederson Unbiased combinations of nonanalog Monte Carlo techniques and fair games. Nuclear Science and Engineering 110: Bucklew, J. A Introduction to rare event simulation. New York: Springer-Verlag. Cancela, H., G. Rubino, and B. Tuffin MTTF estimation by Monte Carlo methods using Markov models. Monte Carlo Methods and Applications 8 (4): Cérou, F., and A. Guyader. 2005, October. Adaptive multilevel splitting for rare event analysis. Technical Report 5710, INRIA. Cérou, F., F. LeGland, P. Del Moral, and P. Lezaud Limit theorems for the multilevel splitting algorithm in the simulation of rare events. In Proceedings of the 2005 Winter Simulation Conference, ed. F. B. A. M. E. Kuhl, N. M. Steiger and J. A. Joines, Del Moral, P Feynman-Kac formulae. genealogical and interacting particle systems with applications. Probability and its Applications. New York: Springer. Ermakov, S. M., and V. B. Melas Design and analysis of simulation experiments. Dordrecht, The Netherlands: Kluwer Academic. Fox, B. L Strategies for quasi-monte Carlo. Boston, MA: Kluwer Academic. Garvels, M. J. J The splitting method in rare event simulation. Ph. D. thesis, Faculty of mathematical Science, University of Twente, The Netherlands. Garvels, M. J. J., and D. P. Kroese A comparison of RESTART implementations. In Proceedings of the 1998 Winter Simulation Conference, : IEEE Press. Garvels, M. J. J., D. P. Kroese, and J.-K. C. W. Van Om-

Rare Events, Splitting, and Quasi-Monte Carlo

Rare Events, Splitting, and Quasi-Monte Carlo Rare Events, Splitting, and Quasi-Monte Carlo PIERRE L ECUYER and VALÉRIE DEMERS Université de Montréal, Canada and BRUNO TUFFIN IRISA INRIA, Rennes, France In the context of rare-event simulation, splitting

More information

Introduction to rare event simulation

Introduction to rare event simulation 1/35 rare event simulation (some works with colleagues H. Cancela, V. Demers, P. L Ecuyer, G. Rubino) INRIA - Centre Bretagne Atlantique, Rennes Aussois, June 2008, AEP9 Outline 2/35 1 2 3 4 5 Introduction:

More information

Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space

Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Pierre L Ecuyer 1, Christian Lécot 2, and Bruno Tuffin 3 1 Département d informatique et de recherche opérationnelle,

More information

ASYMPTOTIC ROBUSTNESS OF ESTIMATORS IN RARE-EVENT SIMULATION. Bruno Tuffin. IRISA/INRIA, Campus de Beaulieu Rennes Cedex, France

ASYMPTOTIC ROBUSTNESS OF ESTIMATORS IN RARE-EVENT SIMULATION. Bruno Tuffin. IRISA/INRIA, Campus de Beaulieu Rennes Cedex, France Submitted to: Proceedings of the 2007 INFORMS Simulation Society Research Workshop ASYMPTOTIC ROBUSTNESS OF ESTIMATORS IN RARE-EVENT SIMULATION Jose H. Blanchet Department of Statistics Harvard University,

More information

Some recent improvements to importance splitting

Some recent improvements to importance splitting Some recent improvements to importance splitting F. Cerou IRISA / INRIA Campus de Beaulieu 35042 RENNES Cédex, France Frederic.Cerou@inria.fr F. Le Gland IRISA / INRIA Campus de Beaulieu 35042 RENNES Cédex,

More information

Resolution-Stationary Random Number Generators

Resolution-Stationary Random Number Generators Resolution-Stationary Random Number Generators Francois Panneton Caisse Centrale Desjardins, 1 Complexe Desjardins, bureau 2822 Montral (Québec), H5B 1B3, Canada Pierre L Ecuyer Département d Informatique

More information

THE STATIC network reliability problem, which consists

THE STATIC network reliability problem, which consists 1 Approximate Zero-Variance Importance Sampling for Static Network Reliability Estimation Pierre L Ecuyer, Gerardo Rubino, Samira Saggadi and Bruno Tuffin Abstract We propose a new Monte Carlo method,

More information

Static Network Reliability Estimation Via Generalized Splitting

Static Network Reliability Estimation Via Generalized Splitting Static Network Reliability Estimation Via Generalized Splitting Zdravko I. Botev, Pierre L Ecuyer DIRO, Université de Montreal, C.P. 6128, Succ. Centre-Ville, Montréal (Québec), H3C 3J7, CANADA, {botev@iro.umontreal.ca,

More information

Transient behaviour in highly dependable Markovian systems: new regimes, many paths.

Transient behaviour in highly dependable Markovian systems: new regimes, many paths. Transient behaviour in highly dependable Markovian systems: new regimes, many paths. Daniël Reijsbergen Pieter-Tjerk de Boer Werner Scheinhardt University of Twente RESIM, June 22nd, 2010 Outline Problem

More information

Importance splitting for rare event simulation

Importance splitting for rare event simulation Importance splitting for rare event simulation F. Cerou Frederic.Cerou@inria.fr Inria Rennes Bretagne Atlantique Simulation of hybrid dynamical systems and applications to molecular dynamics September

More information

Robustesse des techniques de Monte Carlo dans l analyse d événements rares

Robustesse des techniques de Monte Carlo dans l analyse d événements rares Institut National de Recherche en Informatique et Automatique Institut de Recherche en Informatique et Systèmes Aléatoires Robustesse des techniques de Monte Carlo dans l analyse d événements rares H.

More information

Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement

Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement Wyean Chan DIRO, Université de Montréal, C.P. 6128, Succ. Centre-Ville, Montréal (Québec), H3C 3J7, CANADA,

More information

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,

More information

On the estimation of the mean time to failure by simulation

On the estimation of the mean time to failure by simulation On the estimation of the mean time to failure by simulation Peter Glynn, Marvin Nakayama, Bruno Tuffin To cite this version: Peter Glynn, Marvin Nakayama, Bruno Tuffin. On the estimation of the mean time

More information

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,

More information

2 P. L'Ecuyer and R. Simard otherwise perform well in the spectral test, fail this independence test in a decisive way. LCGs with multipliers that hav

2 P. L'Ecuyer and R. Simard otherwise perform well in the spectral test, fail this independence test in a decisive way. LCGs with multipliers that hav Beware of Linear Congruential Generators with Multipliers of the form a = 2 q 2 r Pierre L'Ecuyer and Richard Simard Linear congruential random number generators with Mersenne prime modulus and multipliers

More information

On Derivative Estimation of the Mean Time to Failure in Simulations of Highly Reliable Markovian Systems

On Derivative Estimation of the Mean Time to Failure in Simulations of Highly Reliable Markovian Systems On Derivative Estimation of the Mean Time to Failure in Simulations of Highly Reliable Markovian Systems Marvin K. Nakayama Department of Computer and Information Science New Jersey Institute of Technology

More information

Construction of Equidistributed Generators Based on Linear Recurrences Modulo 2

Construction of Equidistributed Generators Based on Linear Recurrences Modulo 2 Construction of Equidistributed Generators Based on Linear Recurrences Modulo 2 Pierre L Ecuyer and François Panneton Département d informatique et de recherche opérationnelle Université de Montréal C.P.

More information

Asymptotic Robustness of Estimators in Rare-Event Simulation

Asymptotic Robustness of Estimators in Rare-Event Simulation Asymptotic Robustness of Estimators in Rare-Event Simulation PIERRE L ECUYER, Université de Montreal, Canada JOSE H. BLANCHET, Harvard University, USA BRUNO TUFFIN, IRISA INRIA, Rennes, France, and PETER

More information

A Central Limit Theorem for Fleming-Viot Particle Systems Application to the Adaptive Multilevel Splitting Algorithm

A Central Limit Theorem for Fleming-Viot Particle Systems Application to the Adaptive Multilevel Splitting Algorithm A Central Limit Theorem for Fleming-Viot Particle Systems Application to the Algorithm F. Cérou 1,2 B. Delyon 2 A. Guyader 3 M. Rousset 1,2 1 Inria Rennes Bretagne Atlantique 2 IRMAR, Université de Rennes

More information

14.1 Finding frequent elements in stream

14.1 Finding frequent elements in stream Chapter 14 Streaming Data Model 14.1 Finding frequent elements in stream A very useful statistics for many applications is to keep track of elements that occur more frequently. It can come in many flavours

More information

COMBINATION OF CONDITIONAL MONTE CARLO AND APPROXIMATE ZERO-VARIANCE IMPORTANCE SAMPLING FOR NETWORK RELIABILITY ESTIMATION

COMBINATION OF CONDITIONAL MONTE CARLO AND APPROXIMATE ZERO-VARIANCE IMPORTANCE SAMPLING FOR NETWORK RELIABILITY ESTIMATION Proceedings of the 2010 Winter Simulation Conference B. Johansson, S. Jain, J. Montoya-Torres, J. Hugan, and E. Yücesan, eds. COMBINATION OF CONDITIONAL MONTE CARLO AND APPROXIMATE ZERO-VARIANCE IMPORTANCE

More information

Efficient Simulation of a Tandem Queue with Server Slow-down

Efficient Simulation of a Tandem Queue with Server Slow-down Efficient Simulation of a Tandem Queue with Server Slow-down D.I. Miretskiy, W.R.W. Scheinhardt, M.R.H. Mandjes Abstract Tandem Jackson networks, and more sophisticated variants, have found widespread

More information

14 Random Variables and Simulation

14 Random Variables and Simulation 14 Random Variables and Simulation In this lecture note we consider the relationship between random variables and simulation models. Random variables play two important roles in simulation models. We assume

More information

Lecture 9 Classification of States

Lecture 9 Classification of States Lecture 9: Classification of States of 27 Course: M32K Intro to Stochastic Processes Term: Fall 204 Instructor: Gordan Zitkovic Lecture 9 Classification of States There will be a lot of definitions and

More information

Average Reward Parameters

Average Reward Parameters Simulation-Based Optimization of Markov Reward Processes: Implementation Issues Peter Marbach 2 John N. Tsitsiklis 3 Abstract We consider discrete time, nite state space Markov reward processes which depend

More information

Session-Based Queueing Systems

Session-Based Queueing Systems Session-Based Queueing Systems Modelling, Simulation, and Approximation Jeroen Horters Supervisor VU: Sandjai Bhulai Executive Summary Companies often offer services that require multiple steps on the

More information

Asymptotics for Polling Models with Limited Service Policies

Asymptotics for Polling Models with Limited Service Policies Asymptotics for Polling Models with Limited Service Policies Woojin Chang School of Industrial and Systems Engineering Georgia Institute of Technology Atlanta, GA 30332-0205 USA Douglas G. Down Department

More information

An overview of importance splitting for rare event simulation

An overview of importance splitting for rare event simulation Home Search Collections Journals About Contact us My IOPscience An overview of importance splitting for rare event simulation This article has been downloaded from IOPscience. Please scroll down to see

More information

Worst case analysis for a general class of on-line lot-sizing heuristics

Worst case analysis for a general class of on-line lot-sizing heuristics Worst case analysis for a general class of on-line lot-sizing heuristics Wilco van den Heuvel a, Albert P.M. Wagelmans a a Econometric Institute and Erasmus Research Institute of Management, Erasmus University

More information

Stochastic Enumeration Method for Counting Trees

Stochastic Enumeration Method for Counting Trees Stochastic Enumeration Method for Counting Trees Slava Vaisman (Joint work with Dirk P. Kroese) University of Queensland r.vaisman@uq.edu.au January 11, 2015 Slava Vaisman (UQ) Stochastic enumeration January

More information

The square root rule for adaptive importance sampling

The square root rule for adaptive importance sampling The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated

More information

On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules

On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules Vasile Sinescu and Pierre L Ecuyer Abstract We examine the question of constructing shifted lattice rules of rank one with

More information

SOLUTIONS IEOR 3106: Second Midterm Exam, Chapters 5-6, November 8, 2012

SOLUTIONS IEOR 3106: Second Midterm Exam, Chapters 5-6, November 8, 2012 SOLUTIONS IEOR 3106: Second Midterm Exam, Chapters 5-6, November 8, 2012 This exam is closed book. YOU NEED TO SHOW YOUR WORK. Honor Code: Students are expected to behave honorably, following the accepted

More information

HITTING TIME IN AN ERLANG LOSS SYSTEM

HITTING TIME IN AN ERLANG LOSS SYSTEM Probability in the Engineering and Informational Sciences, 16, 2002, 167 184+ Printed in the U+S+A+ HITTING TIME IN AN ERLANG LOSS SYSTEM SHELDON M. ROSS Department of Industrial Engineering and Operations

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

The Wright-Fisher Model and Genetic Drift

The Wright-Fisher Model and Genetic Drift The Wright-Fisher Model and Genetic Drift January 22, 2015 1 1 Hardy-Weinberg Equilibrium Our goal is to understand the dynamics of allele and genotype frequencies in an infinite, randomlymating population

More information

Nets and filters (are better than sequences)

Nets and filters (are better than sequences) Nets and filters (are better than sequences) Contents 1 Motivation 2 2 More implications we wish would reverse 2 3 Nets 4 4 Subnets 6 5 Filters 9 6 The connection between nets and filters 12 7 The payoff

More information

Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3. A renewal process is a generalization of the Poisson point process.

Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3. A renewal process is a generalization of the Poisson point process. Renewal Processes Wednesday, December 16, 2015 1:02 PM Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3 A renewal process is a generalization of the Poisson point process. The Poisson point process is completely

More information

4.5 Applications of Congruences

4.5 Applications of Congruences 4.5 Applications of Congruences 287 66. Find all solutions of the congruence x 2 16 (mod 105). [Hint: Find the solutions of this congruence modulo 3, modulo 5, and modulo 7, and then use the Chinese remainder

More information

Bounded Normal Approximation in Highly Reliable Markovian Systems

Bounded Normal Approximation in Highly Reliable Markovian Systems Bounded Normal Approximation in Highly Reliable Markovian Systems Bruno Tuffin To cite this version: Bruno Tuffin. Bounded Normal Approximation in Highly Reliable Markovian Systems. [Research Report] RR-3020,

More information

Introduction to Rare Event Simulation

Introduction to Rare Event Simulation Introduction to Rare Event Simulation Brown University: Summer School on Rare Event Simulation Jose Blanchet Columbia University. Department of Statistics, Department of IEOR. Blanchet (Columbia) 1 / 31

More information

On the inefficiency of state-independent importance sampling in the presence of heavy tails

On the inefficiency of state-independent importance sampling in the presence of heavy tails Operations Research Letters 35 (2007) 251 260 Operations Research Letters www.elsevier.com/locate/orl On the inefficiency of state-independent importance sampling in the presence of heavy tails Achal Bassamboo

More information

Proceedings of the 2015 Winter Simulation Conference L. Yilmaz, W. K. V. Chan, I. Moon, T. M. K. Roeder, C. Macal, and M. D. Rossetti, eds.

Proceedings of the 2015 Winter Simulation Conference L. Yilmaz, W. K. V. Chan, I. Moon, T. M. K. Roeder, C. Macal, and M. D. Rossetti, eds. Proceedings of the 2015 Winter Simulation Conference L. Yilmaz, W. K. V. Chan, I. Moon, T. M. K. Roeder, C. Macal, and M. D. Rossetti, eds. ON THE ROBUSTNESS OF FISHMAN S BOUND-BASED METHOD FOR THE NETWORK

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.262 Discrete Stochastic Processes Midterm Quiz April 6, 2010 There are 5 questions, each with several parts.

More information

the time it takes until a radioactive substance undergoes a decay

the time it takes until a radioactive substance undergoes a decay 1 Probabilities 1.1 Experiments with randomness Wewillusethetermexperimentinaverygeneralwaytorefertosomeprocess that produces a random outcome. Examples: (Ask class for some first) Here are some discrete

More information

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 7 Sequential Monte Carlo methods III 7 April 2017 Computer Intensive Methods (1) Plan of today s lecture

More information

INTRODUCTION TO MARKOV CHAIN MONTE CARLO

INTRODUCTION TO MARKOV CHAIN MONTE CARLO INTRODUCTION TO MARKOV CHAIN MONTE CARLO 1. Introduction: MCMC In its simplest incarnation, the Monte Carlo method is nothing more than a computerbased exploitation of the Law of Large Numbers to estimate

More information

Monte Carlo Methods for Computation and Optimization (048715)

Monte Carlo Methods for Computation and Optimization (048715) Technion Department of Electrical Engineering Monte Carlo Methods for Computation and Optimization (048715) Lecture Notes Prof. Nahum Shimkin Spring 2015 i PREFACE These lecture notes are intended for

More information

1 Gambler s Ruin Problem

1 Gambler s Ruin Problem 1 Gambler s Ruin Problem Consider a gambler who starts with an initial fortune of $1 and then on each successive gamble either wins $1 or loses $1 independent of the past with probabilities p and q = 1

More information

We are going to discuss what it means for a sequence to converge in three stages: First, we define what it means for a sequence to converge to zero

We are going to discuss what it means for a sequence to converge in three stages: First, we define what it means for a sequence to converge to zero Chapter Limits of Sequences Calculus Student: lim s n = 0 means the s n are getting closer and closer to zero but never gets there. Instructor: ARGHHHHH! Exercise. Think of a better response for the instructor.

More information

Bounded relative error and Vanishing relative error in Monte Carlo evaluation of static Network Reliability measures

Bounded relative error and Vanishing relative error in Monte Carlo evaluation of static Network Reliability measures Bounded relative error and Vanishing relative error in Monte Carlo evaluation of static Network Reliability measures Héctor Cancela Depto. de Investigación Operativa Instituto de Computación, Facultad

More information

Cover Page. The handle holds various files of this Leiden University dissertation

Cover Page. The handle  holds various files of this Leiden University dissertation Cover Page The handle http://hdl.handle.net/1887/39637 holds various files of this Leiden University dissertation Author: Smit, Laurens Title: Steady-state analysis of large scale systems : the successive

More information

1 Introduction (January 21)

1 Introduction (January 21) CS 97: Concrete Models of Computation Spring Introduction (January ). Deterministic Complexity Consider a monotonically nondecreasing function f : {,,..., n} {, }, where f() = and f(n) =. We call f a step

More information

Rare-event Simulation Techniques: An Introduction and Recent Advances

Rare-event Simulation Techniques: An Introduction and Recent Advances Rare-event Simulation Techniques: An Introduction and Recent Advances S. Juneja Tata Institute of Fundamental Research, India juneja@tifr.res.in P. Shahabuddin Columbia University perwez@ieor.columbia.edu

More information

Rare Event Simulation using Monte Carlo Methods

Rare Event Simulation using Monte Carlo Methods Rare Event Simulation using Monte Carlo Methods Edited by Gerardo Rubino And Bruno Tuffin INRIA, Rennes, France A John Wiley and Sons, Ltd., Publication Rare Event Simulation using Monte Carlo Methods

More information

11. Point Detector Tally

11. Point Detector Tally and suppose that we want half the source to be in the range 005 < E < 1 and the other half to be in the range 1 < E < 20 Then the input is SPn -5 a SIn 005 1 20 SBn C 0 5 1 MCNP breaks up the function

More information

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 43, NO. 5, MAY

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 43, NO. 5, MAY IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 43, NO. 5, MAY 1998 631 Centralized and Decentralized Asynchronous Optimization of Stochastic Discrete-Event Systems Felisa J. Vázquez-Abad, Christos G. Cassandras,

More information

A Note on Auxiliary Particle Filters

A Note on Auxiliary Particle Filters A Note on Auxiliary Particle Filters Adam M. Johansen a,, Arnaud Doucet b a Department of Mathematics, University of Bristol, UK b Departments of Statistics & Computer Science, University of British Columbia,

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

Stability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk

Stability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk Stability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk ANSAPW University of Queensland 8-11 July, 2013 1 Outline (I) Fluid

More information

Chapter 11 Advanced Topic Stochastic Processes

Chapter 11 Advanced Topic Stochastic Processes Chapter 11 Advanced Topic Stochastic Processes CHAPTER OUTLINE Section 1 Simple Random Walk Section 2 Markov Chains Section 3 Markov Chain Monte Carlo Section 4 Martingales Section 5 Brownian Motion Section

More information

Solving the Poisson Disorder Problem

Solving the Poisson Disorder Problem Advances in Finance and Stochastics: Essays in Honour of Dieter Sondermann, Springer-Verlag, 22, (295-32) Research Report No. 49, 2, Dept. Theoret. Statist. Aarhus Solving the Poisson Disorder Problem

More information

Combining multiple surrogate models to accelerate failure probability estimation with expensive high-fidelity models

Combining multiple surrogate models to accelerate failure probability estimation with expensive high-fidelity models Combining multiple surrogate models to accelerate failure probability estimation with expensive high-fidelity models Benjamin Peherstorfer a,, Boris Kramer a, Karen Willcox a a Department of Aeronautics

More information

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash Equilibrium Price of Stability Coping With NP-Hardness

More information

Nonuniform Random Variate Generation

Nonuniform Random Variate Generation Nonuniform Random Variate Generation 1 Suppose we have a generator of i.i.d. U(0, 1) r.v. s. We want to generate r.v. s from other distributions, such as normal, Weibull, Poisson, etc. Or other random

More information

Quick Sort Notes , Spring 2010

Quick Sort Notes , Spring 2010 Quick Sort Notes 18.310, Spring 2010 0.1 Randomized Median Finding In a previous lecture, we discussed the problem of finding the median of a list of m elements, or more generally the element of rank m.

More information

The concentration of the chromatic number of random graphs

The concentration of the chromatic number of random graphs The concentration of the chromatic number of random graphs Noga Alon Michael Krivelevich Abstract We prove that for every constant δ > 0 the chromatic number of the random graph G(n, p) with p = n 1/2

More information

Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement

Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement Submitted to imanufacturing & Service Operations Management manuscript MSOM-11-370.R3 Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement (Authors names blinded

More information

Lecture - 02 Rules for Pinch Design Method (PEM) - Part 02

Lecture - 02 Rules for Pinch Design Method (PEM) - Part 02 Process Integration Prof. Bikash Mohanty Department of Chemical Engineering Indian Institute of Technology, Roorkee Module - 05 Pinch Design Method for HEN synthesis Lecture - 02 Rules for Pinch Design

More information

However, reliability analysis is not limited to calculation of the probability of failure.

However, reliability analysis is not limited to calculation of the probability of failure. Probabilistic Analysis probabilistic analysis methods, including the first and second-order reliability methods, Monte Carlo simulation, Importance sampling, Latin Hypercube sampling, and stochastic expansions

More information

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time

More information

IMPORTANCE SAMPLING SIMULATIONS OF PHASE-TYPE QUEUES

IMPORTANCE SAMPLING SIMULATIONS OF PHASE-TYPE QUEUES Proceedings of the 29 Winter Simulation Conference M. D. Rossetti, R. R. Hill, B. Johansson, A. Dunkin, and R. G. Ingalls, eds. IMPORTANCE SAMPLING SIMULATIONS OF PHASE-TYPE QUEUES Poul E. Heegaard Department

More information

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes Ellida M. Khazen * 13395 Coppermine Rd. Apartment 410 Herndon VA 20171 USA Abstract

More information

Sequential Decisions

Sequential Decisions Sequential Decisions A Basic Theorem of (Bayesian) Expected Utility Theory: If you can postpone a terminal decision in order to observe, cost free, an experiment whose outcome might change your terminal

More information

Introduction to Probability

Introduction to Probability LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute

More information

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Eco517 Fall 2004 C. Sims MIDTERM EXAM Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering

More information

Performance of DS-CDMA Systems With Optimal Hard-Decision Parallel Interference Cancellation

Performance of DS-CDMA Systems With Optimal Hard-Decision Parallel Interference Cancellation 2918 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 49, NO. 11, NOVEMBER 2003 Performance of DS-CDMA Systems With Optimal Hard-Decision Parallel Interference Cancellation Remco van der Hofstad Marten J.

More information

1 Maintaining a Dictionary

1 Maintaining a Dictionary 15-451/651: Design & Analysis of Algorithms February 1, 2016 Lecture #7: Hashing last changed: January 29, 2016 Hashing is a great practical tool, with an interesting and subtle theory too. In addition

More information

A Hostile model for network reliability analysis

A Hostile model for network reliability analysis Croatian Operational Research Review 489 CRORR 8(2017), 489 498 A Hostile model for network reliability analysis Daniel Lena 1, Franco Robledo 1 and Pablo Romero 1, 1 Facultad de Ingenería, Universidad

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

A new condition based maintenance model with random improvements on the system after maintenance actions: Optimizing by monte carlo simulation

A new condition based maintenance model with random improvements on the system after maintenance actions: Optimizing by monte carlo simulation ISSN 1 746-7233, England, UK World Journal of Modelling and Simulation Vol. 4 (2008) No. 3, pp. 230-236 A new condition based maintenance model with random improvements on the system after maintenance

More information

Lecture 1: Brief Review on Stochastic Processes

Lecture 1: Brief Review on Stochastic Processes Lecture 1: Brief Review on Stochastic Processes A stochastic process is a collection of random variables {X t (s) : t T, s S}, where T is some index set and S is the common sample space of the random variables.

More information

A Combined RESTART - Cross Entropy Method for Rare Event Estimation with Applications to ATM Networks

A Combined RESTART - Cross Entropy Method for Rare Event Estimation with Applications to ATM Networks A Combined RESTART - Cross Entropy Method for Rare Event Estimation with Applications to ATM Networks M.J.J. Garvels and R.Y. Rubinstein October 9, 2008 Faculty of Mathematical Sciences, University of

More information

Combining Shared Coin Algorithms

Combining Shared Coin Algorithms Combining Shared Coin Algorithms James Aspnes Hagit Attiya Keren Censor Abstract This paper shows that shared coin algorithms can be combined to optimize several complexity measures, even in the presence

More information

Aditya Bhaskara CS 5968/6968, Lecture 1: Introduction and Review 12 January 2016

Aditya Bhaskara CS 5968/6968, Lecture 1: Introduction and Review 12 January 2016 Lecture 1: Introduction and Review We begin with a short introduction to the course, and logistics. We then survey some basics about approximation algorithms and probability. We also introduce some of

More information

Asymptotic redundancy and prolixity

Asymptotic redundancy and prolixity Asymptotic redundancy and prolixity Yuval Dagan, Yuval Filmus, and Shay Moran April 6, 2017 Abstract Gallager (1978) considered the worst-case redundancy of Huffman codes as the maximum probability tends

More information

Exercises Stochastic Performance Modelling. Hamilton Institute, Summer 2010

Exercises Stochastic Performance Modelling. Hamilton Institute, Summer 2010 Exercises Stochastic Performance Modelling Hamilton Institute, Summer Instruction Exercise Let X be a non-negative random variable with E[X ]

More information

NEW ESTIMATORS FOR PARALLEL STEADY-STATE SIMULATIONS

NEW ESTIMATORS FOR PARALLEL STEADY-STATE SIMULATIONS roceedings of the 2009 Winter Simulation Conference M. D. Rossetti, R. R. Hill, B. Johansson, A. Dunkin, and R. G. Ingalls, eds. NEW ESTIMATORS FOR ARALLEL STEADY-STATE SIMULATIONS Ming-hua Hsieh Department

More information

NUMBERS WITH INTEGER COMPLEXITY CLOSE TO THE LOWER BOUND

NUMBERS WITH INTEGER COMPLEXITY CLOSE TO THE LOWER BOUND #A1 INTEGERS 12A (2012): John Selfridge Memorial Issue NUMBERS WITH INTEGER COMPLEXITY CLOSE TO THE LOWER BOUND Harry Altman Department of Mathematics, University of Michigan, Ann Arbor, Michigan haltman@umich.edu

More information

Physics 509: Bootstrap and Robust Parameter Estimation

Physics 509: Bootstrap and Robust Parameter Estimation Physics 509: Bootstrap and Robust Parameter Estimation Scott Oser Lecture #20 Physics 509 1 Nonparametric parameter estimation Question: what error estimate should you assign to the slope and intercept

More information

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( )

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( ) Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr Pr = Pr Pr Pr() Pr Pr. We are given three coins and are told that two of the coins are fair and the

More information

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A ) 6. Brownian Motion. stochastic process can be thought of in one of many equivalent ways. We can begin with an underlying probability space (Ω, Σ, P) and a real valued stochastic process can be defined

More information

AN EFFICIENT IMPORTANCE SAMPLING METHOD FOR RARE EVENT SIMULATION IN LARGE SCALE TANDEM NETWORKS

AN EFFICIENT IMPORTANCE SAMPLING METHOD FOR RARE EVENT SIMULATION IN LARGE SCALE TANDEM NETWORKS Proceedings of the Winter Simulation Conference E. Yücesan, C.-H. Chen, J. L. Snowdon, and J. M. Charnes, eds. AN EFFICIENT IMPORTANCE SAMPLING METHOD FOR RARE EVENT SIMULATION IN LARGE SCALE TANDEM NETWORKS

More information

Geometric and probabilistic approaches towards Conflict Prediction

Geometric and probabilistic approaches towards Conflict Prediction Geometric and probabilistic approaches towards Conflict Prediction G.J. Bakker, H.J. Kremer and H.A.P. Blom Nationaal Lucht- en Ruimtevaartlaboratorium National Aerospace Laboratory NLR Geometric and probabilistic

More information

Lecture 1 : Data Compression and Entropy

Lecture 1 : Data Compression and Entropy CPS290: Algorithmic Foundations of Data Science January 8, 207 Lecture : Data Compression and Entropy Lecturer: Kamesh Munagala Scribe: Kamesh Munagala In this lecture, we will study a simple model for

More information

Testing Problems with Sub-Learning Sample Complexity

Testing Problems with Sub-Learning Sample Complexity Testing Problems with Sub-Learning Sample Complexity Michael Kearns AT&T Labs Research 180 Park Avenue Florham Park, NJ, 07932 mkearns@researchattcom Dana Ron Laboratory for Computer Science, MIT 545 Technology

More information

Stochastic Renewal Processes in Structural Reliability Analysis:

Stochastic Renewal Processes in Structural Reliability Analysis: Stochastic Renewal Processes in Structural Reliability Analysis: An Overview of Models and Applications Professor and Industrial Research Chair Department of Civil and Environmental Engineering University

More information